Computer-aided Testing: Assessment of Automatic Item Generation to Create Multiple Choice Test Items

نویسندگان

  • M. Panczyk
  • J. Belowska
  • J. Gotlib
چکیده

Introduction: Computer-aided testing is an alternative to traditional paper-and-pen tests. Despite the fact that computer-based examination methods are well-established in academic centres, continuous monitoring of their quality is still needed. In the academic year 2013/14, doctoral degree students of the Medical Faculty at Medical University of Warsaw (MUW) took part in the "Reliability in research" course for the first time conducted in the form of blended-learning. To be awarded a credit for the course, the students were supposed to pass a final test of the e-learning course that was made available on the Moodle platform. Aim of study: Evaluation of usefulness of automatically generated computer-interactive multiple choice test for assessment of achievements of blended-learning students. Materials and Methods: A total of 96 PhD students, including 45 first year (25 physicians) and 51 second year (24 physicians) students of a doctoral degree course. It was assumed that both groups of students are comparable and may constitute control groups for each other. The e-learning test results comprised a total of 43 multiple-choice questions (four options to choose from) and were subgrouped into the following categories: (1) Ethical aspects of scientific unreliability, (2) Scientific misconduct, (3) Copyright and research activity, (4) Conflict of interest in research, (5) Rules of “Good Research Practice”. A test set was generated individually for each student out of all questions from the database. The easiness of particular versions and the frequency of using particular questions from the entire pool of questions as well as within the thematic areas were compared to assess the quality of particular question sets. The significance of differences in results was assessed and the mean time necessary for completing the test in both groups was evaluated. Non-parametric Mann-Whitney U test was used for analysis. For all analyses, the a priori level of significance was 0.05. Results: The questions included in automatically generated tests reflected the proportion of questions within the thematic subgroups. Deviation from representativity within the fields was not larger than 1.5% and frequency of using test questions ranged between 1.35 and 3.13% (mean: 2.33% ± 0.45). Total test easiness was high and amounted to 0.854 (0.755 – 1.000), and both groups of students did not differ significantly with respect to this (P> 0.05). No significant differences were also found with respect to the time spent on completing the test by first and second year students (460.9 s ± 124.260 versus 436.9 s ± 135.974, P> 0.05). Conclusions: Reliable computer-based examination methods are used to meet the requirement of uniform rules and criteria of assessment of students' achievements. Automatic generation of question sets with the use of the Moodle platform tools may ensure fair and unbiased assessment of educational progress. The quality of computer-aided testing is comparable to that of traditional paper-and-pen test with reference to assessing the achievement of selected outcomes of education.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Evaluating the Quality of Multilingual Items Generated Using Automatic Processes: Preliminary Results from a Reliability Study

Introduction The traditional approach to test item translation is an effortful, time-consuming process conducted by bilingual or multilingual content specialists. One important problem that arises when content specialists perform translations is the introduction of subjectivity into the process. For instance, Hambleton (1993) reported that when a content specialist is told that another speciali...

متن کامل

Establishing Item Uniqueness for Automatic Item Generation

Objective The growing popularity of Automatic Item Generation (AIG) can be attributed to the increasing demand for the production of large pools of operational test items. AIG is an algorithmic way of generating assessment tasks which combines cognitive theories, psychometric practices and computer technologies. The outcome of this algorithmic transcriptions of assessment task is 1986). The ite...

متن کامل

Partial Knowledge in Multiple-Choice Testing

The intent of this study was to discover the nature of (partial) knowledge as estimated by the multiple-choice (MC) test method. An MC test of vocabulary, including 20 items, was given to 10 participants. Each examinee was required to think aloud while focusing on each item before and while making a response. After each test taker was done with each item, s/he was ...

متن کامل

Methods for Creating and Evaluating the Item Model Structure Used In Automatic Item Generation

represents a relatively new but rapidly evolving research area where cognitive theories, computer technologies, and psychometric practices are used to generate items. In its most ambitious form, AIG can be described as the process of using models to generate statistically calibrated items with the aid of computer technology. Significant developments in AIG research and practice have occurred in...

متن کامل

Utility of Complex Alternatives in Multiple-Choice Items: The Case of All of the Above

This study investigated the utility of all of the above (AOTA) as a test option in multiple-choice items. It aimed at estimating item fit, item difficulty, item discrimination, and guess factor of such a choice. Five reading passages of the Key English Test (KET, 2010) were adapted. The test was reconstructed in 2 parallel forms: Test 1 did not include the abovementioned alternative, whereas Te...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2014